agent environment
AgentRefine: Enhancing Agent Generalization through Refinement Tuning
Fu, Dayuan, He, Keqing, Wang, Yejie, Hong, Wentao, Gongque, Zhuoma, Zeng, Weihao, Wang, Wei, Wang, Jingang, Cai, Xunliang, Xu, Weiran
Large Language Model (LLM) based agents have proved their ability to perform complex tasks like humans. However, there is still a large gap between opensourced LLMs and commercial models like the GPT series. In this paper, we focus on improving the agent generalization capabilities of LLMs via instruction tuning. We first observe that the existing agent training corpus exhibits satisfactory results on held-in evaluation sets but fails to generalize to held-out sets. These agenttuning works face severe formatting errors and are frequently stuck in the same mistake for a long while. We analyze that the poor generalization ability comes from overfitting to several manual agent environments and a lack of adaptation to new situations. They struggle with the wrong action steps and can not learn from the experience but just memorize existing observation-action relations. Inspired by the insight, we propose a novel AgentRefine framework for agent-tuning. The core idea is to enable the model to learn to correct its mistakes via observation in the trajectory. Specifically, we propose an agent synthesis framework to encompass a diverse array of environments and tasks and prompt a strong LLM to refine its error action according to the environment feedback. AgentRefine significantly outperforms state-of-the-art agent-tuning work in terms of generalization ability on diverse agent tasks. It also has better robustness facing perturbation and can generate diversified thought in inference. Our findings establish the correlation between agent generalization and self-refinement and provide a new paradigm for future research. Plenty of agent projects such as AutoGPT (Sig), GPT-Engineer (gpt), and BabyAGI (yoh) have employed LLMs as the core controllers, showing potential for practical applications. Recently, open-sourced LLMs (Dubey et al., 2024; been trained on Held-in task.
Evaluating Cultural and Social Awareness of LLM Web Agents
Qiu, Haoyi, Fabbri, Alexander R., Agarwal, Divyansh, Huang, Kung-Hsiang, Tan, Sarah, Peng, Nanyun, Wu, Chien-Sheng
As large language models (LLMs) expand into performing as agents for real-world applications beyond traditional NLP tasks, evaluating their robustness becomes increasingly important. However, existing benchmarks often overlook critical dimensions like cultural and social awareness. To address these, we introduce CASA, a benchmark designed to assess LLM agents' sensitivity to cultural and social norms across two web-based tasks: online shopping and social discussion forums. Our approach evaluates LLM agents' ability to detect and appropriately respond to norm-violating user queries and observations. Furthermore, we propose a comprehensive evaluation framework that measures awareness coverage, helpfulness in managing user queries, and the violation rate when facing misleading web content. Experiments show that current LLMs perform significantly better in non-agent than in web-based agent environments, with agents achieving less than 10% awareness coverage and over 40% violation rates. To improve performance, we explore two methods: prompting and fine-tuning, and find that combining both methods can offer complementary advantages -- fine-tuning on culture-specific datasets significantly enhances the agents' ability to generalize across different regions, while prompting boosts the agents' ability to navigate complex tasks. These findings highlight the importance of constantly benchmarking LLM agents' cultural and social awareness during the development cycle.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- Asia > China (0.05)
- Asia > Middle East > Iran (0.05)
- (21 more...)
- Research Report (0.64)
- Workflow (0.46)
- Law (1.00)
- Government > Immigration & Customs (0.67)
- Information Technology > Services > e-Commerce Services (0.36)
Understanding Agent Environment in AI - KDnuggets
Before starting the article, it is important to understand what an agent in AI is. The agent is basically an entity that helps the AI, machine learning, or deep reinforcement learning to make a decision or trigger the AI to make a decision. In terms of software, it is defined as the entity which can take decisions and can make different decisions on the basis of changes in the environment, or after getting input from the external environment. In simpler words, the quick agent perceives external change and acts against it the better the results obtained from the model. Hence the role of the agent is always very important in artificial intelligence, machine learning, and deep learning.